5 research outputs found

    CheXpert: A Large Chest Radiograph Dataset with Uncertainty Labels and Expert Comparison

    Full text link
    Large, labeled datasets have driven deep learning methods to achieve expert-level performance on a variety of medical imaging tasks. We present CheXpert, a large dataset that contains 224,316 chest radiographs of 65,240 patients. We design a labeler to automatically detect the presence of 14 observations in radiology reports, capturing uncertainties inherent in radiograph interpretation. We investigate different approaches to using the uncertainty labels for training convolutional neural networks that output the probability of these observations given the available frontal and lateral radiographs. On a validation set of 200 chest radiographic studies which were manually annotated by 3 board-certified radiologists, we find that different uncertainty approaches are useful for different pathologies. We then evaluate our best model on a test set composed of 500 chest radiographic studies annotated by a consensus of 5 board-certified radiologists, and compare the performance of our model to that of 3 additional radiologists in the detection of 5 selected pathologies. On Cardiomegaly, Edema, and Pleural Effusion, the model ROC and PR curves lie above all 3 radiologist operating points. We release the dataset to the public as a standard benchmark to evaluate performance of chest radiograph interpretation models. The dataset is freely available at https://stanfordmlgroup.github.io/competitions/chexpert .Comment: Published in AAAI 201

    Deep COVID DeteCT: an international experience on COVID-19 lung detection and prognosis using chest CT

    Get PDF
    The Coronavirus disease 2019 (COVID-19) presents open questions in how we clinically diagnose and assess disease course. Recently, chest computed tomography (CT) has shown utility for COVID-19 diagnosis. In this study, we developed Deep COVID DeteCT (DCD), a deep learning convolutional neural network (CNN) that uses the entire chest CT volume to automatically predict COVID-19 (COVID+) from non-COVID-19 (COVID−) pneumonia and normal controls. We discuss training strategies and differences in performance across 13 international institutions and 8 countries. The inclusion of non-China sites in training significantly improved classification performance with area under the curve (AUCs) and accuracies above 0.8 on most test sites. Furthermore, using available follow-up scans, we investigate methods to track patient disease course and predict prognosis

    Brain Tumor Radiogenomic Classification of O-Methylguanine-DNA Methyltransferase Promoter Methylation in Malignant Gliomas-Based Transfer Learning

    No full text
    Artificial Intelligence (AI) is the subject of a challenge and attention in the field of oncology and raises many promises for preventive diagnosis, but also fears, some of which are based on highly speculative visions for the classification and detection of tumors. A brain tumor that is malignant is a life-threatening disorder. Glioblastoma is the most prevalent kind of adult brain cancer and the 1 with the poorest prognosis, with a median survival time of less than a year. The presence of O 6 -methylguanine-DNA methyltransferase (MGMT) promoter methylation, a particular genetic sequence seen in tumors, has been proven to be a positive prognostic indicator and a significant predictor of recurrence. This strong revival of interest in AI is modeled in particular to major technological advances which have significantly increased the performance of the predicted model for medical decision support. Establishing reliable forecasts remains a significant challenge for electronic health records (EHRs). By enhancing clinical practice, precision medicine promises to improve healthcare delivery. The goal is to produce improved prognosis, diagnosis, and therapy through evidence-based sub stratification of patients, transforming established clinical pathways to optimize care for each patient’s individual requirements. The abundance of today’s healthcare data, dubbed “big data,” provides great resources for new knowledge discovery, potentially advancing precision treatment. The latter necessitates multidisciplinary initiatives that will use the knowledge, skills, and medical data of newly established organizations with diverse backgrounds and expertise. The aim of this paper is to use magnetic resonance imaging (MRI) images to train and evaluate your model to detect the presence of MGMT promoter methylation in this competition to predict the genetic subtype of glioblastoma based transfer learning. Our objective is to emphasize the basic problems in the developing disciplines of radiomics and radiogenomics, as well as to illustrate the computational challenges from the perspective of big data analytics

    Differentiation of benign and malignant lymph nodes in pediatric patients on ferumoxytol-enhanced PET/MRI

    No full text
    The composition of lymph nodes in pediatric patients is different from that in adults. Most notably, normal lymph nodes in children contain less macrophages. Therefore, previously described biodistributions of iron oxide nanoparticles in benign and malignant lymph nodes of adult patients may not apply to children. The purpose of our study was to evaluate if the iron supplement ferumoxytol improves the differentiation of benign and malignant lymph nodes in pediatric cancer patients on F-18-FDG PET/MRI. Methods: We conducted a prospective clinical trial from May 2015 to December 2018 to investigate the value of ferumoxytol nanoparticles for staging of children with cancer with F-18-FDG PET/MRI. Ferumoxytol is an FDA-approved iron supplement for the treatment of anemia and has been used off-label as an MRI contrast agent in this study. Forty-two children (7-18 years, 29 male, 13 female) received a F-18-FDG PET/MRI at 2 (n=20) or 24 hours (h) (n=22) after intravenous injection of ferumoxytol (dose 5 mg Fe/kg). The morphology of benign and malignant lymph nodes on ferumoxytol-enhanced T2-FSE sequences at 2 and 24 h were compared using a linear regression analysis. In addition, ADCmean-values, SUV-ratio (SUVmax lesion/SUVmean liver) and R2*-relaxation rate of benign and malignant lymph nodes were compared with a Mann-Whitney-U test. The accuracy of different criteria was assessed with a receiver operating characteristics (ROC) curve. Follow-up imaging for at least 6 months served as the standard of reference. Results: We examined a total of 613 lymph nodes, of which 464 (75.7%) were benign and 149 (24.3%) were malignant. On ferumoxytol-enhanced T2-FSE images, benign lymph nodes showed a hypointense hilum and hyperintense parenchyma, while malignant lymph nodes showed no discernible hilum. This pattern was not significantly different at 2 h and 24 h postcontrast (p=0.82). Benign and malignant lymph nodes showed significantly different ferumoxytol enhancement patterns, ADCmean values of 1578 and 852 x10(-6) mm(2)/s, mean SUV-ratios of 0.5 and 2.8, and mean R2*-relaxation rate of 127.8 and 84.4 Hertz (Hz), respectively (all p90%) of ADCmean, SUV-ratio, pattern, and R2* measurements for the characterization of benign and malignant lymph nodes in children. Ferumoxytol nanoparticle accumulation at the hilum can be used to diagnose a benign lymph node. In the future, the delivery of clinically applicable nanoparticles to the hilum of benign lymph nodes could be harnessed to deliver theranostic drugs for immune cell priming

    Deep learning for chest radiograph diagnosis: A retrospective comparison of the CheXNeXt algorithm to practicing radiologists.

    No full text
    BackgroundChest radiograph interpretation is critical for the detection of thoracic diseases, including tuberculosis and lung cancer, which affect millions of people worldwide each year. This time-consuming task typically requires expert radiologists to read the images, leading to fatigue-based diagnostic error and lack of diagnostic expertise in areas of the world where radiologists are not available. Recently, deep learning approaches have been able to achieve expert-level performance in medical image interpretation tasks, powered by large network architectures and fueled by the emergence of large labeled datasets. The purpose of this study is to investigate the performance of a deep learning algorithm on the detection of pathologies in chest radiographs compared with practicing radiologists.Methods and findingsWe developed CheXNeXt, a convolutional neural network to concurrently detect the presence of 14 different pathologies, including pneumonia, pleural effusion, pulmonary masses, and nodules in frontal-view chest radiographs. CheXNeXt was trained and internally validated on the ChestX-ray8 dataset, with a held-out validation set consisting of 420 images, sampled to contain at least 50 cases of each of the original pathology labels. On this validation set, the majority vote of a panel of 3 board-certified cardiothoracic specialist radiologists served as reference standard. We compared CheXNeXt's discriminative performance on the validation set to the performance of 9 radiologists using the area under the receiver operating characteristic curve (AUC). The radiologists included 6 board-certified radiologists (average experience 12 years, range 4-28 years) and 3 senior radiology residents, from 3 academic institutions. We found that CheXNeXt achieved radiologist-level performance on 11 pathologies and did not achieve radiologist-level performance on 3 pathologies. The radiologists achieved statistically significantly higher AUC performance on cardiomegaly, emphysema, and hiatal hernia, with AUCs of 0.888 (95% confidence interval [CI] 0.863-0.910), 0.911 (95% CI 0.866-0.947), and 0.985 (95% CI 0.974-0.991), respectively, whereas CheXNeXt's AUCs were 0.831 (95% CI 0.790-0.870), 0.704 (95% CI 0.567-0.833), and 0.851 (95% CI 0.785-0.909), respectively. CheXNeXt performed better than radiologists in detecting atelectasis, with an AUC of 0.862 (95% CI 0.825-0.895), statistically significantly higher than radiologists' AUC of 0.808 (95% CI 0.777-0.838); there were no statistically significant differences in AUCs for the other 10 pathologies. The average time to interpret the 420 images in the validation set was substantially longer for the radiologists (240 minutes) than for CheXNeXt (1.5 minutes). The main limitations of our study are that neither CheXNeXt nor the radiologists were permitted to use patient history or review prior examinations and that evaluation was limited to a dataset from a single institution.ConclusionsIn this study, we developed and validated a deep learning algorithm that classified clinically important abnormalities in chest radiographs at a performance level comparable to practicing radiologists. Once tested prospectively in clinical settings, the algorithm could have the potential to expand patient access to chest radiograph diagnostics
    corecore